Goto

Collaborating Authors

 example show




Paying Attention to Facts: Quantifying the Knowledge Capacity of Attention Layers

Wong, Liang Ze

arXiv.org Artificial Intelligence

In this paper, we investigate the ability of single-layer attention-only transformers (i.e. attention layers) to memorize facts contained in databases from a linear-algebraic perspective. We associate with each database a 3-tensor, propose the rank of this tensor as a measure of the size of the database, and provide bounds on the rank in terms of properties of the database. We also define a 3-tensor corresponding to an attention layer, and empirically demonstrate the relationship between its rank and database rank on a dataset of toy models and random databases. By highlighting the roles played by the value-output and query-key weights, and the effects of argmax and softmax on rank, our results shed light on the `additive motif' of factual recall in transformers, while also suggesting a way of increasing layer capacity without increasing the number of parameters.


Exploring the Limits of Semantic Image Compression at Micro-bits per Pixel

Dotzel, Jordan, Kotb, Bahaa, Dotzel, James, Abdelfattah, Mohamed, Zhang, Zhiru

arXiv.org Artificial Intelligence

Traditional methods, such as JPEG, perform image compression by operating on structural information, such as pixel values or frequency content. These methods are effective to bitrates around one bit per pixel (bpp) and higher at standard image sizes. In contrast, text-based semantic compression directly stores concepts and their relationships using natural language, which has evolved with humans to efficiently represent these salient concepts. These methods can operate at extremely low bitrates by disregarding structural information like location, size, and orientation. In this work, we use GPT-4V and DALL-E3 from OpenAI to explore the quality-compression frontier for image compression and identify the limitations of current technology. We push semantic compression as low as 100 $\mu$bpp (up to $10,000\times$ smaller than JPEG) by introducing an iterative reflection process to improve the decoded image. We further hypothesize this 100 $\mu$bpp level represents a soft limit on semantic compression at standard image resolutions.


Moving Forward by Moving Backward: Embedding Action Impact over Action Semantics

Zeng, Kuo-Hao, Weihs, Luca, Mottaghi, Roozbeh, Farhadi, Ali

arXiv.org Artificial Intelligence

A common assumption when training embodied agents is that the impact of taking an action is stable; for instance, executing the "move ahead" action will always move the agent forward by a fixed distance, perhaps with some small amount of actuator-induced noise. This assumption is limiting; an agent may encounter settings that dramatically alter the impact of actions: a move ahead action on a wet floor may send the agent twice as far as it expects and using the same action with a broken wheel might transform the expected translation into a rotation. Instead of relying that the impact of an action stably reflects its pre-defined semantic meaning, we propose to model the impact of actions on-the-fly using latent embeddings. We evaluate our AAP on two challenging visual navigation tasks in the AI2-THOR and Habitat environments and show that our AAP is highly performant even when faced, at inference-time with missing actions and, previously unseen, perturbed action space. Moreover, we observe significant improvement in robustness against these actions when evaluating in real-world scenarios. Humans show a remarkable capacity for planning when faced with substantially constrained or augmented means by which they may interact with their environment. For instance, a human who begins to walk on ice will readily shorten their stride to prevent slipping. Likewise, a human will spare little mental effort in deciding to exert more force to lift their hand when it is weighed down by groceries. Even in these mundane tasks, we see that the effect of a humans' actions can have significantly different outcomes depending on the setting: there is no predefined one-to-one mapping between actions and their impact. The same is true for embodied agents where something as simple as attempting to moving forward can result in radically different outcomes depending on the load the agent carries, the presence of surface debris, and the maintenance level of the agent's actuators (e.g., are any wheels broken?). We call this the action-stability assumption (AS assumption).


GitHub - aws/amazon-sagemaker-examples: Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.

#artificialintelligence

These examples show you how to use SageMaker Processing jobs to run data processing workloads. These examples show you how to use SageMaker Pipelines to create, automate and manage end-to-end Machine Learning workflows. These examples show you how to train and host in pre-built deep learning framework containers using the SageMaker Python SDK. These examples show you how to build Machine Learning models with frameworks like Apache Spark or Scikit-learn using SageMaker Python SDK. These examples show how to use Amazon SageMaker for model training, hosting, and inference through Apache Spark using SageMaker Spark.


Announcing support for extracting data from identity documents using Amazon Textract

#artificialintelligence

Creating efficiencies in your business is at the top of your list. You want your employees to be more productive, have them focus on high impact tasks, or find ways to implement better processes to improve the outcomes to your customers. There are various ways to solve this problem, and more companies are turning to artificial intelligence (AI) and machine learning (ML) to help. In the financial services sector, there is the creation of new accounts online, or in healthcare there are new digital platforms to schedule and manage appointments, which require users to fill out forms. These can be error prone, time consuming, and certainly improved upon.


8 Revolutionary Applications of Machine Learning in Practice

#artificialintelligence

Machine learning (ML) is an innovative tool that advances technology in every industry around the world. Due to its constant learning and evolution, the algorithms are able to adapt based on success and failure. Then, they can help people in daily life. Machine learning mimics the human brain. It entails deep learning from its neural networks, natural language processing (NLP), and constant changes based on incoming information. Of course, these algorithms aren't perfect, but they become more refined with every interaction.


10 Azure ML Code Examples Every Cloud AI Developer Should Know

#artificialintelligence

TLDR; The Azure ML Python SDK enables Data scientists, AI engineers,and MLOps developers to be productive in the cloud. This post highlights 10 examples every cloud AI developer should know, to be successful with Azure ML. If you are new to Azure you can get a free subscription using the link below. The scripts in this example are used to classify iris flower images to build a machine learning model based on scikit-learn's iris dataset the code can easily be adapted to any scikit-learn estimator. This example shows you how to run your TensorFlow training scripts at scale using Azure Machine Learning's TensorFlow estimator class.


Deep Learning Examples: R2020a Edition

#artificialintelligence

With two releases every year, you may find it challenging to keep up with the latest features.* In fact, some people who work here feel the same way! This release, I asked the Product Managers about the new features related to deep learning that they think you should know about in release 20a. Here are their responses: Deep Learning Starting with Deep Learning Toolbox, there are three new features to get excited about in 20a. Experiment Manager (new) - A new app that keeps track all conditions when training neural networks.